2026-04-10 18:34:21,013 [ 180162 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2026-04-10 18:34:21,013 [ 180162 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:79, check_args_and_update_paths) 2026-04-10 18:34:21,013 [ 180162 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:90, check_args_and_update_paths) 2026-04-10 18:34:21,013 [ 180162 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:92, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_9p9r35 --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel0_1.jsonl --report-log-exclude-logs-on-passed-tests test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated 'test_cow_policy/test.py::test_cow_policy[cow_policy_multi_disk]' 'test_cow_policy/test.py::test_cow_policy[cow_policy_multi_volume]' -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [6 items] scheduling tests via LoadFileScheduling test_cow_policy/test.py::test_cow_policy[cow_policy_multi_disk] test_attach_partition_using_copy/test.py::test_all_replicated [gw1] [ 16%] FAILED test_cow_policy/test.py::test_cow_policy[cow_policy_multi_disk] test_cow_policy/test.py::test_cow_policy[cow_policy_multi_volume] [gw0] [ 33%] FAILED test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree [gw1] [ 50%] FAILED test_cow_policy/test.py::test_cow_policy[cow_policy_multi_volume] [gw0] [ 66%] FAILED test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk [gw0] [ 83%] FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated [gw0] [100%] FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated =================================== FAILURES =================================== ____________________ test_cow_policy[cow_policy_multi_disk] ____________________ [gw1] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = storage_policy = 'cow_policy_multi_disk' @pytest.mark.parametrize("storage_policy", ["cow_policy_multi_disk", "cow_policy_multi_volume"]) def test_cow_policy(start_cluster, storage_policy): try: > node.query_with_retry( f""" ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS storage_policy = '{storage_policy}' """, timeout=60, retry_count=3, ) test_cow_policy/test.py:24: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = sql = "\n ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7'\n (\n ...R BY (postcode1, postcode2, addr1, addr2)\n SETTINGS storage_policy = 'cow_policy_multi_disk'\n " stdin = None, timeout = 60, settings = None, user = None, password = None database = None, host = None, ignore_error = False, retry_count = 3 sleep_time = 0.5 check_callback = at 0x7fa733e5c040> parse = False def query_with_retry( self, sql, stdin=None, timeout=None, settings=None, user=None, password=None, database=None, host=None, ignore_error=False, retry_count=20, sleep_time=0.5, check_callback=lambda x: True, parse=False, ): # logging.debug(f"Executing query {sql} on {self.name}") result = None exception_msg = "" for i in range(retry_count): try: result = self.query( sql, stdin=stdin, timeout=timeout, settings=settings, user=user, password=password, database=database, host=host, ignore_error=ignore_error, parse=parse, ) if check_callback(result): return result time.sleep(sleep_time) except QueryRuntimeException as ex: exception_msg = f"{type(ex).__name__}: {str(ex)}" # Container is down, this is likely due to server crash. if "No route to host" in str(ex): raise time.sleep(sleep_time) except Exception as ex: # logging.debug("Retry {} got exception {}".format(i + 1, ex)) exception_msg = f"{type(ex).__name__}: {str(ex)}" time.sleep(sleep_time) if result is not None: return result > raise Exception(f"Can't execute query {sql}\n{exception_msg}") E Exception: Can't execute query E ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS storage_policy = 'cow_policy_multi_disk' E E QueryRuntimeException: Client failed! Return code: 198, stderr: Received exception from server (version 25.3.8): E Code: 198. DB::Exception: Received from 172.16.2.2:9000. DB::NetException. DB::NetException: Not found address of host: raw.githubusercontent.com: while loading disk metadata. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:113: Poco::Exception::Exception(String const&, int) @ 0x000000001f81eb40 E 1. ./build_docker/./src/Common/Exception.cpp:108: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000ff71534 E 2. ./src/Common/Exception.h:112: DB::NetException::NetException(int, FormatStringHelperImpl::type>, String const&) @ 0x000000000ff2de19 E 3. ./build_docker/./src/Common/DNSResolver.cpp:113: DB::(anonymous namespace)::hostByName(String const&) @ 0x000000000ff29783 E 4. ./build_docker/./src/Common/DNSResolver.cpp:138: DB::DNSResolver::getResolvedIPAdressessWithFiltering(String const&) @ 0x000000000ff27d53 E 5. ./build_docker/./src/Common/DNSResolver.cpp:256: DB::DNSResolver::resolveIPAddressWithCache(String const&) @ 0x000000000ff28339 E 6. ./build_docker/./src/Common/DNSResolver.cpp:276: DB::DNSResolver::resolveHostAllInOriginOrder(String const&) @ 0x000000000ff28a70 E 7. ./build_docker/./src/Common/HostResolvePool.cpp:54: std::vector> std::__function::__policy_invoker> (String const&)>::__call_impl[abi:ne190107]> (String const&)>>(std::__function::__policy_storage const*, String const&) @ 0x00000000103fb829 E 8. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x00000000103f968f E 9. ./build_docker/./src/Common/HostResolvePool.cpp:66: DB::HostResolver::HostResolver(std::function> (String const&)>&&, String, Poco::Timespan) @ 0x00000000103f949f E 10. ./build_docker/./src/Common/HostResolvePool.cpp:53: DB::HostResolver::HostResolver(String, Poco::Timespan) @ 0x00000000103f913e E 11. ./src/Common/HostResolvePool.h:62: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler::make_shared_enabler(String const&) @ 0x00000000103fe723 E 12. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler> std::allocate_shared[abi:ne190107] DB::HostResolver::create(String const&)::make_shared_enabler, std::allocator DB::HostResolver::create(String const&)::make_shared_enabler>, String const&, 0>(std::allocator DB::HostResolver::create(String const&)::make_shared_enabler> const&, String const&) @ 0x00000000103fe426 E 13. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:851: DB::HostResolversPool::getResolver(String const&) @ 0x00000000103fb6f1 E 14. ./build_docker/./src/Common/HTTPConnectionPool.cpp:671: DB::EndpointConnectionPool::prepareNewConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103edbb6 E 15. ./build_docker/./src/Common/HTTPConnectionPool.cpp:590: DB::EndpointConnectionPool::getConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103ece18 E 16. ./build_docker/./src/IO/HTTPCommon.cpp:63: DB::makeHTTPSession(DB::HTTPConnectionGroupType, Poco::URI const&, DB::ConnectionTimeouts const&, DB::ProxyConfiguration const&, unsigned long*) @ 0x000000001040592d E 17. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:267: DB::ReadWriteBufferFromHTTP::callImpl(Poco::Net::HTTPResponse&, String const&, std::optional const&, bool) const @ 0x0000000013133465 E 18. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:285: DB::ReadWriteBufferFromHTTP::callWithRedirects(Poco::Net::HTTPResponse&, String const&, std::optional const&) @ 0x0000000013133714 E 19. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:408: DB::ReadWriteBufferFromHTTP::initialize() @ 0x00000000131341c0 E 20. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:472: void std::__function::__policy_invoker::__call_impl[abi:ne190107]>(std::__function::__policy_storage const*) @ 0x0000000013137377 E 21. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x0000000013130bd9 E 22. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:465: DB::ReadWriteBufferFromHTTP::nextImpl() @ 0x0000000013135881 E 23. DB::ReadBuffer::next() @ 0x0000000007d5819b E 24. ./src/IO/ReadBuffer.h:96: DB::WebObjectStorage::loadFiles(String const&, std::unique_lock const&) const @ 0x0000000016d85459 E 25. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:225: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d888da E 26. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:185: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d882fa E 27. ./build_docker/./src/Disks/ObjectStorages/Web/MetadataStorageFromStaticFilesWebServer.cpp:106: DB::MetadataStorageFromStaticFilesWebServer::getStorageObjectsIfExist(String const&) const @ 0x0000000016d828a5 E 28. ./build_docker/./src/Disks/ObjectStorages/DiskObjectStorage.cpp:785: DB::DiskObjectStorage::readFileIfExists(String const&, DB::ReadSettings const&, std::optional, std::optional) const @ 0x0000000016ce8e92 E 29. ./build_docker/./src/Storages/MergeTree/MergeTreeData.cpp:380: DB::MergeTreeData::initializeDirectoriesAndFormatVersion(String const&, bool, String const&, bool) @ 0x000000001a7120fd E 30. ./build_docker/./src/Storages/StorageMergeTree.cpp:159: DB::StorageMergeTree::StorageMergeTree(DB::StorageID const&, String const&, DB::StorageInMemoryMetadata const&, DB::LoadingStrictnessLevel, std::shared_ptr, String const&, DB::MergeTreeData::MergingParams const&, std::unique_ptr>) @ 0x000000001abe70e7 E 31. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr std::allocate_shared[abi:ne190107], DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>, 0>(std::allocator const&, DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>&&) @ 0x000000001abe6a04 E . (DNS_ERROR) E (query: ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS storage_policy = 'cow_policy_multi_disk' E ) helpers/cluster.py:3722: Exception ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ------------------------------ Captured log setup ------------------------------ 2026-04-10 18:34:28.021000 [ 625 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.042000 [ 625 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:34:28.042000 [ 625 ] DEBUG : No running containers (conftest.py:95, cleanup_environment) 2026-04-10 18:34:28.042000 [ 625 ] DEBUG : Pruning Docker networks (conftest.py:97, cleanup_environment) 2026-04-10 18:34:28.042000 [ 625 ] DEBUG : Command:[docker network prune --force] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.064000 [ 625 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.067000 [ 625 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:145, run_and_check) 2026-04-10 18:34:28.068000 [ 625 ] INFO : Running tests in /ClickHouse/tests/integration/test_cow_policy/test.py (cluster.py:2748, start) 2026-04-10 18:34:28.068000 [ 625 ] DEBUG : Cluster start called. is_up=False (cluster.py:2755, start) 2026-04-10 18:34:28.087000 [ 625 ] DEBUG : Docker networks for project roottestcowpolicy-gw1 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:34:28.107000 [ 625 ] DEBUG : Docker containers for project roottestcowpolicy-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:34:28.127000 [ 625 ] DEBUG : Docker volumes for project roottestcowpolicy-gw1 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:34:28.127000 [ 625 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 18:34:28.146000 [ 625 ] DEBUG : Docker networks for project roottestcowpolicy-gw1 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:34:28.166000 [ 625 ] DEBUG : Docker containers for project roottestcowpolicy-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:34:28.185000 [ 625 ] DEBUG : Docker volumes for project roottestcowpolicy-gw1 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:34:28.185000 [ 625 ] DEBUG : Command:[docker container list --all --filter name='^/roottestcowpolicy-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.204000 [ 625 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 18:34:28.204000 [ 625 ] DEBUG : No running containers for project: roottestcowpolicy-gw1 (cluster.py:889, cleanup) 2026-04-10 18:34:28.204000 [ 625 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 18:34:28.226000 [ 625 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 18:34:28.226000 [ 625 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.253000 [ 625 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 18:34:28.253000 [ 625 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 18:34:28.254000 [ 625 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 18:34:28.254000 [ 625 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.272000 [ 625 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:34:28.272000 [ 625 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) 2026-04-10 18:34:28.272000 [ 625 ] DEBUG : Setup directory for instance: node (cluster.py:2768, start) 2026-04-10 18:34:28.273000 [ 625 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4638, create_dir) 2026-04-10 18:34:28.273000 [ 625 ] DEBUG : Create directory for common tests configuration (cluster.py:4643, create_dir) 2026-04-10 18:34:28.273000 [ 625 ] DEBUG : Copy common configuration from helpers (cluster.py:4663, create_dir) 2026-04-10 18:34:28.274000 [ 625 ] DEBUG : Generate and write macros file (cluster.py:4715, create_dir) 2026-04-10 18:34:28.275000 [ 625 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_cow_policy/configs/overrides.yaml'] to /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/configs/config.d (cluster.py:4755, create_dir) 2026-04-10 18:34:28.275000 [ 625 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/database (cluster.py:4772, create_dir) 2026-04-10 18:34:28.276000 [ 625 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/logs (cluster.py:4783, create_dir) 2026-04-10 18:34:28.276000 [ 625 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2026-04-10 18:34:28.276000 [ 625 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'LSAN_OPTIONS': 'suppressions=/etc/clickhouse-server/lsan_suppressions.txt', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/.env (cluster.py:96, _create_env_file) 2026-04-10 18:34:28.276000 [ 625 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 18:34:28.276000 [ 625 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 18:34:28.277000 [ 625 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 18:34:28.277000 [ 625 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 18:34:28.287000 [ 625 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2026-04-10 18:34:28.287000 [ 625 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/.env --project-name roottestcowpolicy-gw1 --file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/docker-compose.yml pull] (cluster.py:121, run_and_check) 2026-04-10 18:34:38.731000 [ 625 ] DEBUG : Stderr: node Pulling (cluster.py:147, run_and_check) 2026-04-10 18:34:38.732000 [ 625 ] DEBUG : Stderr: node Pulled (cluster.py:147, run_and_check) 2026-04-10 18:34:38.732000 [ 625 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/.env --project-name roottestcowpolicy-gw1 --file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/docker-compose.yml up -d --no-recreate') (cluster.py:3149, start) 2026-04-10 18:34:38.732000 [ 625 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/.env --project-name roottestcowpolicy-gw1 --file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/docker-compose.yml up -d --no-recreate] (cluster.py:121, run_and_check) 2026-04-10 18:34:39.620000 [ 625 ] DEBUG : Stderr: Network roottestcowpolicy-gw1_default Creating (cluster.py:147, run_and_check) 2026-04-10 18:34:39.621000 [ 625 ] DEBUG : Stderr: Network roottestcowpolicy-gw1_default Created (cluster.py:147, run_and_check) 2026-04-10 18:34:39.621000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Creating (cluster.py:147, run_and_check) 2026-04-10 18:34:39.621000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Created (cluster.py:147, run_and_check) 2026-04-10 18:34:39.621000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Starting (cluster.py:147, run_and_check) 2026-04-10 18:34:39.621000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Started (cluster.py:147, run_and_check) 2026-04-10 18:34:39.621000 [ 625 ] DEBUG : ClickHouse instance created (cluster.py:3157, start) 2026-04-10 18:34:39.621000 [ 625 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2015, get_instance_ip) 2026-04-10 18:34:39.623000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestcowpolicy-gw1-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:39.624000 [ 625 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2025, get_instance_global_ipv6) 2026-04-10 18:34:39.626000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestcowpolicy-gw1-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:39.626000 [ 625 ] DEBUG : Waiting for ClickHouse start in node, ip: 172.16.2.2... (cluster.py:3165, start) 2026-04-10 18:34:39.628000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestcowpolicy-gw1-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:39.630000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:39.733000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:39.836000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:39.940000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.044000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.149000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.253000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.356000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.460000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.564000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.667000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.770000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.873000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:40.976000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.080000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.183000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.287000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.390000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.494000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.597000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.701000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.804000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:41.908000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:42.011000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:42.115000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:42.219000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:42.323000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:42.427000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:42.531000 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/dcde3ea7bfcc261c8f3acc59f320ece03f08603ada7f89a6b66e911c05bff63b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:42.532000 [ 625 ] DEBUG : ClickHouse node started (cluster.py:3169, start) ------------------------------ Captured log call ------------------------------- 2026-04-10 18:34:42.535000 [ 625 ] DEBUG : Executing query ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS storage_policy = 'cow_policy_multi_disk' on node (cluster.py:3658, query) 2026-04-10 18:35:36.527000 [ 625 ] DEBUG : Executing query ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS storage_policy = 'cow_policy_multi_disk' on node (cluster.py:3658, query) 2026-04-10 18:36:31.061000 [ 625 ] DEBUG : Executing query ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS storage_policy = 'cow_policy_multi_disk' on node (cluster.py:3658, query) 2026-04-10 18:37:26.297000 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS uk_price_paid SYNC on node (cluster.py:3658, query) _____________________________ test_all_replicated ______________________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_all_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", True) test_attach_partition_using_copy/test.py:128: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = sql = "\n ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7'\n (\n price UInt32,\n ...disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/')\n " stdin = None, timeout = 60, settings = None, user = None, password = None database = None, host = None, ignore_error = False, retry_count = 3 sleep_time = 0.5 check_callback = at 0x7f1cb8be8040> parse = False def query_with_retry( self, sql, stdin=None, timeout=None, settings=None, user=None, password=None, database=None, host=None, ignore_error=False, retry_count=20, sleep_time=0.5, check_callback=lambda x: True, parse=False, ): # logging.debug(f"Executing query {sql} on {self.name}") result = None exception_msg = "" for i in range(retry_count): try: result = self.query( sql, stdin=stdin, timeout=timeout, settings=settings, user=user, password=password, database=database, host=host, ignore_error=ignore_error, parse=parse, ) if check_callback(result): return result time.sleep(sleep_time) except QueryRuntimeException as ex: exception_msg = f"{type(ex).__name__}: {str(ex)}" # Container is down, this is likely due to server crash. if "No route to host" in str(ex): raise time.sleep(sleep_time) except Exception as ex: # logging.debug("Retry {} got exception {}".format(i + 1, ex)) exception_msg = f"{type(ex).__name__}: {str(ex)}" time.sleep(sleep_time) if result is not None: return result > raise Exception(f"Can't execute query {sql}\n{exception_msg}") E Exception: Can't execute query E ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') E E QueryRuntimeException: Client failed! Return code: 198, stderr: Received exception from server (version 25.3.8): E Code: 198. DB::Exception: Received from 172.16.1.5:9000. DB::NetException. DB::NetException: Not found address of host: raw.githubusercontent.com: while loading disk metadata. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:113: Poco::Exception::Exception(String const&, int) @ 0x000000001f81eb40 E 1. ./build_docker/./src/Common/Exception.cpp:108: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000ff71534 E 2. ./src/Common/Exception.h:112: DB::NetException::NetException(int, FormatStringHelperImpl::type>, String const&) @ 0x000000000ff2de19 E 3. ./build_docker/./src/Common/DNSResolver.cpp:113: DB::(anonymous namespace)::hostByName(String const&) @ 0x000000000ff29783 E 4. ./build_docker/./src/Common/DNSResolver.cpp:138: DB::DNSResolver::getResolvedIPAdressessWithFiltering(String const&) @ 0x000000000ff27d53 E 5. ./build_docker/./src/Common/DNSResolver.cpp:256: DB::DNSResolver::resolveIPAddressWithCache(String const&) @ 0x000000000ff28339 E 6. ./build_docker/./src/Common/DNSResolver.cpp:276: DB::DNSResolver::resolveHostAllInOriginOrder(String const&) @ 0x000000000ff28a70 E 7. ./build_docker/./src/Common/HostResolvePool.cpp:54: std::vector> std::__function::__policy_invoker> (String const&)>::__call_impl[abi:ne190107]> (String const&)>>(std::__function::__policy_storage const*, String const&) @ 0x00000000103fb829 E 8. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x00000000103f968f E 9. ./build_docker/./src/Common/HostResolvePool.cpp:66: DB::HostResolver::HostResolver(std::function> (String const&)>&&, String, Poco::Timespan) @ 0x00000000103f949f E 10. ./build_docker/./src/Common/HostResolvePool.cpp:53: DB::HostResolver::HostResolver(String, Poco::Timespan) @ 0x00000000103f913e E 11. ./src/Common/HostResolvePool.h:62: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler::make_shared_enabler(String const&) @ 0x00000000103fe723 E 12. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler> std::allocate_shared[abi:ne190107] DB::HostResolver::create(String const&)::make_shared_enabler, std::allocator DB::HostResolver::create(String const&)::make_shared_enabler>, String const&, 0>(std::allocator DB::HostResolver::create(String const&)::make_shared_enabler> const&, String const&) @ 0x00000000103fe426 E 13. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:851: DB::HostResolversPool::getResolver(String const&) @ 0x00000000103fb6f1 E 14. ./build_docker/./src/Common/HTTPConnectionPool.cpp:671: DB::EndpointConnectionPool::prepareNewConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103edbb6 E 15. ./build_docker/./src/Common/HTTPConnectionPool.cpp:590: DB::EndpointConnectionPool::getConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103ece18 E 16. ./build_docker/./src/IO/HTTPCommon.cpp:63: DB::makeHTTPSession(DB::HTTPConnectionGroupType, Poco::URI const&, DB::ConnectionTimeouts const&, DB::ProxyConfiguration const&, unsigned long*) @ 0x000000001040592d E 17. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:267: DB::ReadWriteBufferFromHTTP::callImpl(Poco::Net::HTTPResponse&, String const&, std::optional const&, bool) const @ 0x0000000013133465 E 18. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:285: DB::ReadWriteBufferFromHTTP::callWithRedirects(Poco::Net::HTTPResponse&, String const&, std::optional const&) @ 0x0000000013133714 E 19. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:408: DB::ReadWriteBufferFromHTTP::initialize() @ 0x00000000131341c0 E 20. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:472: void std::__function::__policy_invoker::__call_impl[abi:ne190107]>(std::__function::__policy_storage const*) @ 0x0000000013137377 E 21. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x0000000013130bd9 E 22. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:465: DB::ReadWriteBufferFromHTTP::nextImpl() @ 0x0000000013135881 E 23. DB::ReadBuffer::next() @ 0x0000000007d5819b E 24. ./src/IO/ReadBuffer.h:96: DB::WebObjectStorage::loadFiles(String const&, std::unique_lock const&) const @ 0x0000000016d85459 E 25. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:225: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d888da E 26. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:185: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d882fa E 27. ./build_docker/./src/Disks/ObjectStorages/Web/MetadataStorageFromStaticFilesWebServer.cpp:106: DB::MetadataStorageFromStaticFilesWebServer::getStorageObjectsIfExist(String const&) const @ 0x0000000016d828a5 E 28. ./build_docker/./src/Disks/ObjectStorages/DiskObjectStorage.cpp:785: DB::DiskObjectStorage::readFileIfExists(String const&, DB::ReadSettings const&, std::optional, std::optional) const @ 0x0000000016ce8e92 E 29. ./build_docker/./src/Storages/MergeTree/MergeTreeData.cpp:380: DB::MergeTreeData::initializeDirectoriesAndFormatVersion(String const&, bool, String const&, bool) @ 0x000000001a7120fd E 30. ./build_docker/./src/Storages/StorageReplicatedMergeTree.cpp:414: DB::StorageReplicatedMergeTree::StorageReplicatedMergeTree(DB::TableZnodeInfo const&, DB::LoadingStrictnessLevel, DB::StorageID const&, String const&, DB::StorageInMemoryMetadata const&, std::shared_ptr, String const&, DB::MergeTreeData::MergingParams const&, std::unique_ptr>, bool, DB::ZooKeeperRetriesInfo const&) @ 0x000000001a060dd4 E 31. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr std::allocate_shared[abi:ne190107], DB::TableZnodeInfo&, DB::LoadingStrictnessLevel const&, DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>, bool&, DB::ZooKeeperRetriesInfo&, 0>(std::allocator const&, DB::TableZnodeInfo&, DB::LoadingStrictnessLevel const&, DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>&&, bool&, DB::ZooKeeperRetriesInfo&) @ 0x000000001abe6671 E . (DNS_ERROR) E (query: ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') E ) helpers/cluster.py:3722: Exception ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ------------------------------ Captured log setup ------------------------------ 2026-04-10 18:34:28.021000 [ 622 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.042000 [ 622 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:34:28.042000 [ 622 ] DEBUG : No running containers (conftest.py:95, cleanup_environment) 2026-04-10 18:34:28.042000 [ 622 ] DEBUG : Pruning Docker networks (conftest.py:97, cleanup_environment) 2026-04-10 18:34:28.042000 [ 622 ] DEBUG : Command:[docker network prune --force] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.065000 [ 622 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.067000 [ 622 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:145, run_and_check) 2026-04-10 18:34:28.068000 [ 622 ] INFO : Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py (cluster.py:2748, start) 2026-04-10 18:34:28.068000 [ 622 ] DEBUG : Cluster start called. is_up=False (cluster.py:2755, start) 2026-04-10 18:34:28.087000 [ 622 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:34:28.108000 [ 622 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:34:28.129000 [ 622 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:34:28.129000 [ 622 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 18:34:28.149000 [ 622 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:34:28.170000 [ 622 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:34:28.189000 [ 622 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:34:28.190000 [ 622 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.212000 [ 622 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 18:34:28.213000 [ 622 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw0 (cluster.py:889, cleanup) 2026-04-10 18:34:28.213000 [ 622 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 18:34:28.232000 [ 622 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 18:34:28.232000 [ 622 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.250000 [ 622 ] DEBUG : Stderr:Error response from daemon: a prune operation is already running (cluster.py:147, run_and_check) 2026-04-10 18:34:28.250000 [ 622 ] DEBUG : Exitcode:1 (cluster.py:149, run_and_check) 2026-04-10 18:34:28.251000 [ 622 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 18:34:28.251000 [ 622 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:34:28.271000 [ 622 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:34:28.271000 [ 622 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) 2026-04-10 18:34:28.271000 [ 622 ] DEBUG : Setup directory for instance: replica1 (cluster.py:2768, start) 2026-04-10 18:34:28.272000 [ 622 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4638, create_dir) 2026-04-10 18:34:28.272000 [ 622 ] DEBUG : Create directory for common tests configuration (cluster.py:4643, create_dir) 2026-04-10 18:34:28.273000 [ 622 ] DEBUG : Copy common configuration from helpers (cluster.py:4663, create_dir) 2026-04-10 18:34:28.273000 [ 622 ] DEBUG : Generate and write macros file (cluster.py:4715, create_dir) 2026-04-10 18:34:28.274000 [ 622 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/configs/config.d (cluster.py:4755, create_dir) 2026-04-10 18:34:28.274000 [ 622 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/database (cluster.py:4772, create_dir) 2026-04-10 18:34:28.274000 [ 622 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/logs (cluster.py:4783, create_dir) 2026-04-10 18:34:28.274000 [ 622 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2026-04-10 18:34:28.274000 [ 622 ] DEBUG : Setup directory for instance: replica2 (cluster.py:2768, start) 2026-04-10 18:34:28.275000 [ 622 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4638, create_dir) 2026-04-10 18:34:28.275000 [ 622 ] DEBUG : Create directory for common tests configuration (cluster.py:4643, create_dir) 2026-04-10 18:34:28.275000 [ 622 ] DEBUG : Copy common configuration from helpers (cluster.py:4663, create_dir) 2026-04-10 18:34:28.276000 [ 622 ] DEBUG : Generate and write macros file (cluster.py:4715, create_dir) 2026-04-10 18:34:28.276000 [ 622 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/configs/config.d (cluster.py:4755, create_dir) 2026-04-10 18:34:28.276000 [ 622 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/database (cluster.py:4772, create_dir) 2026-04-10 18:34:28.277000 [ 622 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/logs (cluster.py:4783, create_dir) 2026-04-10 18:34:28.277000 [ 622 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2026-04-10 18:34:28.277000 [ 622 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'LSAN_OPTIONS': 'suppressions=/etc/clickhouse-server/lsan_suppressions.txt', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:5ccda723c1fc', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/.env (cluster.py:96, _create_env_file) 2026-04-10 18:34:28.277000 [ 622 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 18:34:28.277000 [ 622 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 18:34:28.278000 [ 622 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 18:34:28.278000 [ 622 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 18:34:28.287000 [ 622 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2026-04-10 18:34:28.288000 [ 622 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/.env --project-name roottestattachpartitionusingcopy-gw0 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/docker-compose.yml pull] (cluster.py:121, run_and_check) 2026-04-10 18:34:38.727000 [ 622 ] DEBUG : Stderr: zoo3 Skipped - Image is already being pulled by replica1 (cluster.py:147, run_and_check) 2026-04-10 18:34:38.727000 [ 622 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by replica1 (cluster.py:147, run_and_check) 2026-04-10 18:34:38.727000 [ 622 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by replica1 (cluster.py:147, run_and_check) 2026-04-10 18:34:38.727000 [ 622 ] DEBUG : Stderr: replica2 Skipped - Image is already being pulled by replica1 (cluster.py:147, run_and_check) 2026-04-10 18:34:38.727000 [ 622 ] DEBUG : Stderr: replica1 Pulling (cluster.py:147, run_and_check) 2026-04-10 18:34:38.728000 [ 622 ] DEBUG : Stderr: replica1 Pulled (cluster.py:147, run_and_check) 2026-04-10 18:34:38.728000 [ 622 ] DEBUG : Setup ZooKeeper (cluster.py:2809, start) 2026-04-10 18:34:38.728000 [ 622 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/keeper3/coordination'] (cluster.py:2810, start) 2026-04-10 18:34:38.729000 [ 622 ] DEBUG : Command:[docker compose --project-name roottestattachpartitionusingcopy-gw0 --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:121, run_and_check) 2026-04-10 18:34:39.405000 [ 622 ] DEBUG : Stderr:time="2026-04-10T18:34:38Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:147, run_and_check) 2026-04-10 18:34:39.405000 [ 622 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw0_default Creating (cluster.py:147, run_and_check) 2026-04-10 18:34:39.405000 [ 622 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw0_default Created (cluster.py:147, run_and_check) 2026-04-10 18:34:39.405000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Creating (cluster.py:147, run_and_check) 2026-04-10 18:34:39.405000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Creating (cluster.py:147, run_and_check) 2026-04-10 18:34:39.405000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Creating (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Created (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Created (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Created (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Starting (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Starting (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Starting (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Started (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Started (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Started (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr:time="2026-04-10T18:34:39Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Stderr:time="2026-04-10T18:34:39Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : Wait ZooKeeper to start (cluster.py:2446, wait_zookeeper_to_start) 2026-04-10 18:34:39.406000 [ 622 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2015, get_instance_ip) 2026-04-10 18:34:39.409000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:39.410000 [ 622 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.1.4, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 18:34:39.411000 [ 622 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 18:34:39.412000 [ 622 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 18:34:39.523000 [ 622 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 18:34:39.524000 [ 622 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 18:34:39.795000 [ 622 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 18:34:39.795000 [ 622 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 18:34:40.425000 [ 622 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 18:34:40.426000 [ 622 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 18:34:41.389000 [ 622 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 18:34:41.390000 [ 622 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 18:34:43.660000 [ 622 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 18:34:43.661000 [ 622 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 18:34:43.670000 [ 622 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 18:34:43.671000 [ 622 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 18:34:43.672000 [ 622 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 18:34:43.672000 [ 622 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 18:34:43.677000 [ 622 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 18:34:43.677000 [ 622 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 18:34:43.677000 [ 622 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 18:34:43.743000 [ 622 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 18:34:43.743000 [ 622 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 18:34:43.743000 [ 622 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2015, get_instance_ip) 2026-04-10 18:34:43.746000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:43.747000 [ 622 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.1.2, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 18:34:43.748000 [ 622 ] INFO : Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 18:34:43.749000 [ 622 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 18:34:43.754000 [ 622 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 18:34:43.754000 [ 622 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 18:34:43.755000 [ 622 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 18:34:43.755000 [ 622 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 18:34:43.759000 [ 622 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 18:34:43.759000 [ 622 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 18:34:43.760000 [ 622 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 18:34:43.860000 [ 622 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 18:34:43.860000 [ 622 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 18:34:43.861000 [ 622 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2015, get_instance_ip) 2026-04-10 18:34:43.864000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:43.864000 [ 622 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.1.3, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 18:34:43.866000 [ 622 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 18:34:43.867000 [ 622 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 18:34:43.876000 [ 622 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 18:34:43.876000 [ 622 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 18:34:43.877000 [ 622 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 18:34:43.878000 [ 622 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 18:34:43.883000 [ 622 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 18:34:43.883000 [ 622 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 18:34:43.883000 [ 622 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 18:34:43.984000 [ 622 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 18:34:43.984000 [ 622 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 18:34:43.984000 [ 622 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2462, wait_zookeeper_nodes_to_start) 2026-04-10 18:34:43.985000 [ 622 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/.env --project-name roottestattachpartitionusingcopy-gw0 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/docker-compose.yml up -d --no-recreate') (cluster.py:3149, start) 2026-04-10 18:34:43.985000 [ 622 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/.env --project-name roottestattachpartitionusingcopy-gw0 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/docker-compose.yml up -d --no-recreate] (cluster.py:121, run_and_check) 2026-04-10 18:34:44.440000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Running (cluster.py:147, run_and_check) 2026-04-10 18:34:44.440000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Running (cluster.py:147, run_and_check) 2026-04-10 18:34:44.440000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Running (cluster.py:147, run_and_check) 2026-04-10 18:34:44.440000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Creating (cluster.py:147, run_and_check) 2026-04-10 18:34:44.440000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Creating (cluster.py:147, run_and_check) 2026-04-10 18:34:44.440000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Created (cluster.py:147, run_and_check) 2026-04-10 18:34:44.440000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Created (cluster.py:147, run_and_check) 2026-04-10 18:34:44.440000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Starting (cluster.py:147, run_and_check) 2026-04-10 18:34:44.441000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Starting (cluster.py:147, run_and_check) 2026-04-10 18:34:44.441000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Started (cluster.py:147, run_and_check) 2026-04-10 18:34:44.441000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Started (cluster.py:147, run_and_check) 2026-04-10 18:34:44.441000 [ 622 ] DEBUG : ClickHouse instance created (cluster.py:3157, start) 2026-04-10 18:34:44.441000 [ 622 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2015, get_instance_ip) 2026-04-10 18:34:44.443000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:44.444000 [ 622 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2025, get_instance_global_ipv6) 2026-04-10 18:34:44.446000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:44.447000 [ 622 ] DEBUG : Waiting for ClickHouse start in replica1, ip: 172.16.1.5... (cluster.py:3165, start) 2026-04-10 18:34:44.449000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:44.451000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:44.555000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:44.658000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:44.762000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:44.866000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:44.970000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.073000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.176000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.279000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.382000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.486000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.590000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.694000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.797000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:45.900000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.004000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.107000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.211000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.314000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.417000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.521000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.625000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.728000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.831000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/981cdfa309a5753023c5178d4e725beade606c23b47e811a5cc04447c29a5987/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.832000 [ 622 ] DEBUG : ClickHouse replica1 started (cluster.py:3169, start) 2026-04-10 18:34:46.832000 [ 622 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2015, get_instance_ip) 2026-04-10 18:34:46.835000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.835000 [ 622 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2025, get_instance_global_ipv6) 2026-04-10 18:34:46.837000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.838000 [ 622 ] DEBUG : Waiting for ClickHouse start in replica2, ip: 172.16.1.6... (cluster.py:3165, start) 2026-04-10 18:34:46.840000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw0-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.842000 [ 622 ] DEBUG : http://localhost:None "GET /v1.46/containers/2afc1380d52aa8ab6fdbf8d6cb8fc8931a9b275ce33c18a3313b26f4f67fbf2b/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:34:46.843000 [ 622 ] DEBUG : ClickHouse replica2 started (cluster.py:3169, start) ------------------------------ Captured log call ------------------------------- 2026-04-10 18:34:46.845000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3658, query) 2026-04-10 18:34:47.111000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3658, query) 2026-04-10 18:34:47.326000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3658, query) 2026-04-10 18:34:47.542000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3658, query) 2026-04-10 18:34:47.758000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) 2026-04-10 18:35:41.694000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) 2026-04-10 18:36:36.228000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) ___________________ test_cow_policy[cow_policy_multi_volume] ___________________ [gw1] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = storage_policy = 'cow_policy_multi_volume' @pytest.mark.parametrize("storage_policy", ["cow_policy_multi_disk", "cow_policy_multi_volume"]) def test_cow_policy(start_cluster, storage_policy): try: > node.query_with_retry( f""" ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS storage_policy = '{storage_policy}' """, timeout=60, retry_count=3, ) test_cow_policy/test.py:24: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = sql = "\n ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7'\n (\n ...BY (postcode1, postcode2, addr1, addr2)\n SETTINGS storage_policy = 'cow_policy_multi_volume'\n " stdin = None, timeout = 60, settings = None, user = None, password = None database = None, host = None, ignore_error = False, retry_count = 3 sleep_time = 0.5 check_callback = at 0x7fa733e5c040> parse = False def query_with_retry( self, sql, stdin=None, timeout=None, settings=None, user=None, password=None, database=None, host=None, ignore_error=False, retry_count=20, sleep_time=0.5, check_callback=lambda x: True, parse=False, ): # logging.debug(f"Executing query {sql} on {self.name}") result = None exception_msg = "" for i in range(retry_count): try: result = self.query( sql, stdin=stdin, timeout=timeout, settings=settings, user=user, password=password, database=database, host=host, ignore_error=ignore_error, parse=parse, ) if check_callback(result): return result time.sleep(sleep_time) except QueryRuntimeException as ex: exception_msg = f"{type(ex).__name__}: {str(ex)}" # Container is down, this is likely due to server crash. if "No route to host" in str(ex): raise time.sleep(sleep_time) except Exception as ex: # logging.debug("Retry {} got exception {}".format(i + 1, ex)) exception_msg = f"{type(ex).__name__}: {str(ex)}" time.sleep(sleep_time) if result is not None: return result > raise Exception(f"Can't execute query {sql}\n{exception_msg}") E Exception: Can't execute query E ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS storage_policy = 'cow_policy_multi_volume' E E QueryRuntimeException: Client failed! Return code: 198, stderr: Received exception from server (version 25.3.8): E Code: 198. DB::Exception: Received from 172.16.2.2:9000. DB::NetException. DB::NetException: Not found address of host: raw.githubusercontent.com: while loading disk metadata. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:113: Poco::Exception::Exception(String const&, int) @ 0x000000001f81eb40 E 1. ./build_docker/./src/Common/Exception.cpp:108: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000ff71534 E 2. ./src/Common/Exception.h:112: DB::NetException::NetException(int, FormatStringHelperImpl::type>, String const&) @ 0x000000000ff2de19 E 3. ./build_docker/./src/Common/DNSResolver.cpp:113: DB::(anonymous namespace)::hostByName(String const&) @ 0x000000000ff29783 E 4. ./build_docker/./src/Common/DNSResolver.cpp:138: DB::DNSResolver::getResolvedIPAdressessWithFiltering(String const&) @ 0x000000000ff27d53 E 5. ./build_docker/./src/Common/DNSResolver.cpp:256: DB::DNSResolver::resolveIPAddressWithCache(String const&) @ 0x000000000ff28339 E 6. ./build_docker/./src/Common/DNSResolver.cpp:276: DB::DNSResolver::resolveHostAllInOriginOrder(String const&) @ 0x000000000ff28a70 E 7. ./build_docker/./src/Common/HostResolvePool.cpp:54: std::vector> std::__function::__policy_invoker> (String const&)>::__call_impl[abi:ne190107]> (String const&)>>(std::__function::__policy_storage const*, String const&) @ 0x00000000103fb829 E 8. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x00000000103f968f E 9. ./build_docker/./src/Common/HostResolvePool.cpp:66: DB::HostResolver::HostResolver(std::function> (String const&)>&&, String, Poco::Timespan) @ 0x00000000103f949f E 10. ./build_docker/./src/Common/HostResolvePool.cpp:53: DB::HostResolver::HostResolver(String, Poco::Timespan) @ 0x00000000103f913e E 11. ./src/Common/HostResolvePool.h:62: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler::make_shared_enabler(String const&) @ 0x00000000103fe723 E 12. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler> std::allocate_shared[abi:ne190107] DB::HostResolver::create(String const&)::make_shared_enabler, std::allocator DB::HostResolver::create(String const&)::make_shared_enabler>, String const&, 0>(std::allocator DB::HostResolver::create(String const&)::make_shared_enabler> const&, String const&) @ 0x00000000103fe426 E 13. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:851: DB::HostResolversPool::getResolver(String const&) @ 0x00000000103fb6f1 E 14. ./build_docker/./src/Common/HTTPConnectionPool.cpp:671: DB::EndpointConnectionPool::prepareNewConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103edbb6 E 15. ./build_docker/./src/Common/HTTPConnectionPool.cpp:590: DB::EndpointConnectionPool::getConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103ece18 E 16. ./build_docker/./src/IO/HTTPCommon.cpp:63: DB::makeHTTPSession(DB::HTTPConnectionGroupType, Poco::URI const&, DB::ConnectionTimeouts const&, DB::ProxyConfiguration const&, unsigned long*) @ 0x000000001040592d E 17. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:267: DB::ReadWriteBufferFromHTTP::callImpl(Poco::Net::HTTPResponse&, String const&, std::optional const&, bool) const @ 0x0000000013133465 E 18. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:285: DB::ReadWriteBufferFromHTTP::callWithRedirects(Poco::Net::HTTPResponse&, String const&, std::optional const&) @ 0x0000000013133714 E 19. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:408: DB::ReadWriteBufferFromHTTP::initialize() @ 0x00000000131341c0 E 20. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:472: void std::__function::__policy_invoker::__call_impl[abi:ne190107]>(std::__function::__policy_storage const*) @ 0x0000000013137377 E 21. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x0000000013130bd9 E 22. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:465: DB::ReadWriteBufferFromHTTP::nextImpl() @ 0x0000000013135881 E 23. DB::ReadBuffer::next() @ 0x0000000007d5819b E 24. ./src/IO/ReadBuffer.h:96: DB::WebObjectStorage::loadFiles(String const&, std::unique_lock const&) const @ 0x0000000016d85459 E 25. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:225: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d888da E 26. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:185: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d882fa E 27. ./build_docker/./src/Disks/ObjectStorages/Web/MetadataStorageFromStaticFilesWebServer.cpp:106: DB::MetadataStorageFromStaticFilesWebServer::getStorageObjectsIfExist(String const&) const @ 0x0000000016d828a5 E 28. ./build_docker/./src/Disks/ObjectStorages/DiskObjectStorage.cpp:785: DB::DiskObjectStorage::readFileIfExists(String const&, DB::ReadSettings const&, std::optional, std::optional) const @ 0x0000000016ce8e92 E 29. ./build_docker/./src/Storages/MergeTree/MergeTreeData.cpp:380: DB::MergeTreeData::initializeDirectoriesAndFormatVersion(String const&, bool, String const&, bool) @ 0x000000001a7120fd E 30. ./build_docker/./src/Storages/StorageMergeTree.cpp:159: DB::StorageMergeTree::StorageMergeTree(DB::StorageID const&, String const&, DB::StorageInMemoryMetadata const&, DB::LoadingStrictnessLevel, std::shared_ptr, String const&, DB::MergeTreeData::MergingParams const&, std::unique_ptr>) @ 0x000000001abe70e7 E 31. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr std::allocate_shared[abi:ne190107], DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>, 0>(std::allocator const&, DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>&&) @ 0x000000001abe6a04 E . (DNS_ERROR) E (query: ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS storage_policy = 'cow_policy_multi_volume' E ) helpers/cluster.py:3722: Exception ------------------------------ Captured log call ------------------------------- 2026-04-10 18:37:26.779000 [ 625 ] DEBUG : Executing query ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS storage_policy = 'cow_policy_multi_volume' on node (cluster.py:3658, query) 2026-04-10 18:38:23.876000 [ 625 ] DEBUG : Executing query ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS storage_policy = 'cow_policy_multi_volume' on node (cluster.py:3658, query) 2026-04-10 18:39:21.679000 [ 625 ] DEBUG : Executing query ATTACH TABLE uk_price_paid UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS storage_policy = 'cow_policy_multi_volume' on node (cluster.py:3658, query) 2026-04-10 18:40:16.319000 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS uk_price_paid SYNC on node (cluster.py:3658, query) ---------------------------- Captured log teardown ----------------------------- 2026-04-10 18:40:16.615000 [ 625 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/.env --project-name roottestcowpolicy-gw1 --file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/docker-compose.yml stop --timeout 20] (cluster.py:121, run_and_check) 2026-04-10 18:40:22.780000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:40:22.780000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:40:22.780000 [ 625 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2026-04-10 18:40:22.790000 [ 625 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/.env --project-name roottestcowpolicy-gw1 --file /ClickHouse/tests/integration/test_cow_policy/_instances-1-gw1/node/docker-compose.yml down --volumes] (cluster.py:121, run_and_check) 2026-04-10 18:40:23.302000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:40:23.302000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:40:23.302000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Removing (cluster.py:147, run_and_check) 2026-04-10 18:40:23.302000 [ 625 ] DEBUG : Stderr: Container roottestcowpolicy-gw1-node-1 Removed (cluster.py:147, run_and_check) 2026-04-10 18:40:23.302000 [ 625 ] DEBUG : Stderr: Network roottestcowpolicy-gw1_default Removing (cluster.py:147, run_and_check) 2026-04-10 18:40:23.302000 [ 625 ] DEBUG : Stderr: Network roottestcowpolicy-gw1_default Removed (cluster.py:147, run_and_check) 2026-04-10 18:40:23.303000 [ 625 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 18:40:23.323000 [ 625 ] DEBUG : Docker networks for project roottestcowpolicy-gw1 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:40:23.340000 [ 625 ] DEBUG : Docker containers for project roottestcowpolicy-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:40:23.357000 [ 625 ] DEBUG : Docker volumes for project roottestcowpolicy-gw1 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:40:23.358000 [ 625 ] DEBUG : Command:[docker container list --all --filter name='^/roottestcowpolicy-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 18:40:23.375000 [ 625 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 18:40:23.375000 [ 625 ] DEBUG : No running containers for project: roottestcowpolicy-gw1 (cluster.py:889, cleanup) 2026-04-10 18:40:23.375000 [ 625 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 18:40:23.393000 [ 625 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 18:40:23.393000 [ 625 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 18:40:23.421000 [ 625 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 18:40:23.421000 [ 625 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 18:40:23.421000 [ 625 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 18:40:23.421000 [ 625 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:40:23.440000 [ 625 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:40:23.440000 [ 625 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) _____________________________ test_both_mergetree ______________________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_both_mergetree(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:106: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = sql = "\n ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7'\n (\n price UInt32,\n ...disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/')\n " stdin = None, timeout = 60, settings = None, user = None, password = None database = None, host = None, ignore_error = False, retry_count = 3 sleep_time = 0.5 check_callback = at 0x7f1cb8be8040> parse = False def query_with_retry( self, sql, stdin=None, timeout=None, settings=None, user=None, password=None, database=None, host=None, ignore_error=False, retry_count=20, sleep_time=0.5, check_callback=lambda x: True, parse=False, ): # logging.debug(f"Executing query {sql} on {self.name}") result = None exception_msg = "" for i in range(retry_count): try: result = self.query( sql, stdin=stdin, timeout=timeout, settings=settings, user=user, password=password, database=database, host=host, ignore_error=ignore_error, parse=parse, ) if check_callback(result): return result time.sleep(sleep_time) except QueryRuntimeException as ex: exception_msg = f"{type(ex).__name__}: {str(ex)}" # Container is down, this is likely due to server crash. if "No route to host" in str(ex): raise time.sleep(sleep_time) except Exception as ex: # logging.debug("Retry {} got exception {}".format(i + 1, ex)) exception_msg = f"{type(ex).__name__}: {str(ex)}" time.sleep(sleep_time) if result is not None: return result > raise Exception(f"Can't execute query {sql}\n{exception_msg}") E Exception: Can't execute query E ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree() E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') E E QueryRuntimeException: Client failed! Return code: 198, stderr: Received exception from server (version 25.3.8): E Code: 198. DB::Exception: Received from 172.16.1.5:9000. DB::NetException. DB::NetException: Not found address of host: raw.githubusercontent.com: while loading disk metadata. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:113: Poco::Exception::Exception(String const&, int) @ 0x000000001f81eb40 E 1. ./build_docker/./src/Common/Exception.cpp:108: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000ff71534 E 2. ./src/Common/Exception.h:112: DB::NetException::NetException(int, FormatStringHelperImpl::type>, String const&) @ 0x000000000ff2de19 E 3. ./build_docker/./src/Common/DNSResolver.cpp:113: DB::(anonymous namespace)::hostByName(String const&) @ 0x000000000ff29783 E 4. ./build_docker/./src/Common/DNSResolver.cpp:138: DB::DNSResolver::getResolvedIPAdressessWithFiltering(String const&) @ 0x000000000ff27d53 E 5. ./build_docker/./src/Common/DNSResolver.cpp:256: DB::DNSResolver::resolveIPAddressWithCache(String const&) @ 0x000000000ff28339 E 6. ./build_docker/./src/Common/DNSResolver.cpp:276: DB::DNSResolver::resolveHostAllInOriginOrder(String const&) @ 0x000000000ff28a70 E 7. ./build_docker/./src/Common/HostResolvePool.cpp:54: std::vector> std::__function::__policy_invoker> (String const&)>::__call_impl[abi:ne190107]> (String const&)>>(std::__function::__policy_storage const*, String const&) @ 0x00000000103fb829 E 8. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x00000000103f968f E 9. ./build_docker/./src/Common/HostResolvePool.cpp:66: DB::HostResolver::HostResolver(std::function> (String const&)>&&, String, Poco::Timespan) @ 0x00000000103f949f E 10. ./build_docker/./src/Common/HostResolvePool.cpp:53: DB::HostResolver::HostResolver(String, Poco::Timespan) @ 0x00000000103f913e E 11. ./src/Common/HostResolvePool.h:62: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler::make_shared_enabler(String const&) @ 0x00000000103fe723 E 12. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler> std::allocate_shared[abi:ne190107] DB::HostResolver::create(String const&)::make_shared_enabler, std::allocator DB::HostResolver::create(String const&)::make_shared_enabler>, String const&, 0>(std::allocator DB::HostResolver::create(String const&)::make_shared_enabler> const&, String const&) @ 0x00000000103fe426 E 13. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:851: DB::HostResolversPool::getResolver(String const&) @ 0x00000000103fb6f1 E 14. ./build_docker/./src/Common/HTTPConnectionPool.cpp:671: DB::EndpointConnectionPool::prepareNewConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103edbb6 E 15. ./build_docker/./src/Common/HTTPConnectionPool.cpp:590: DB::EndpointConnectionPool::getConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103ece18 E 16. ./build_docker/./src/IO/HTTPCommon.cpp:63: DB::makeHTTPSession(DB::HTTPConnectionGroupType, Poco::URI const&, DB::ConnectionTimeouts const&, DB::ProxyConfiguration const&, unsigned long*) @ 0x000000001040592d E 17. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:267: DB::ReadWriteBufferFromHTTP::callImpl(Poco::Net::HTTPResponse&, String const&, std::optional const&, bool) const @ 0x0000000013133465 E 18. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:285: DB::ReadWriteBufferFromHTTP::callWithRedirects(Poco::Net::HTTPResponse&, String const&, std::optional const&) @ 0x0000000013133714 E 19. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:408: DB::ReadWriteBufferFromHTTP::initialize() @ 0x00000000131341c0 E 20. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:472: void std::__function::__policy_invoker::__call_impl[abi:ne190107]>(std::__function::__policy_storage const*) @ 0x0000000013137377 E 21. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x0000000013130bd9 E 22. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:465: DB::ReadWriteBufferFromHTTP::nextImpl() @ 0x0000000013135881 E 23. DB::ReadBuffer::next() @ 0x0000000007d5819b E 24. ./src/IO/ReadBuffer.h:96: DB::WebObjectStorage::loadFiles(String const&, std::unique_lock const&) const @ 0x0000000016d85459 E 25. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:225: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d888da E 26. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:185: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d882fa E 27. ./build_docker/./src/Disks/ObjectStorages/Web/MetadataStorageFromStaticFilesWebServer.cpp:106: DB::MetadataStorageFromStaticFilesWebServer::getStorageObjectsIfExist(String const&) const @ 0x0000000016d828a5 E 28. ./build_docker/./src/Disks/ObjectStorages/DiskObjectStorage.cpp:785: DB::DiskObjectStorage::readFileIfExists(String const&, DB::ReadSettings const&, std::optional, std::optional) const @ 0x0000000016ce8e92 E 29. ./build_docker/./src/Storages/MergeTree/MergeTreeData.cpp:380: DB::MergeTreeData::initializeDirectoriesAndFormatVersion(String const&, bool, String const&, bool) @ 0x000000001a7120fd E 30. ./build_docker/./src/Storages/StorageMergeTree.cpp:159: DB::StorageMergeTree::StorageMergeTree(DB::StorageID const&, String const&, DB::StorageInMemoryMetadata const&, DB::LoadingStrictnessLevel, std::shared_ptr, String const&, DB::MergeTreeData::MergingParams const&, std::unique_ptr>) @ 0x000000001abe70e7 E 31. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr std::allocate_shared[abi:ne190107], DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>, 0>(std::allocator const&, DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>&&) @ 0x000000001abe6a04 E . (DNS_ERROR) E (query: ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree() E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') E ) helpers/cluster.py:3722: Exception ------------------------------ Captured log call ------------------------------- 2026-04-10 18:37:31.669000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3658, query) 2026-04-10 18:37:31.885000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3658, query) 2026-04-10 18:37:32.101000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3658, query) 2026-04-10 18:37:32.317000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3658, query) 2026-04-10 18:37:32.583000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) 2026-04-10 18:38:28.582000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) 2026-04-10 18:39:26.381000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) _______________________ test_not_work_on_different_disk ________________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_not_work_on_different_disk(start_cluster): cleanup([replica1, replica2]) # Replace and move should not work on replace > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:199: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = sql = "\n ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7'\n (\n price UInt32,\n ...disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/')\n " stdin = None, timeout = 60, settings = None, user = None, password = None database = None, host = None, ignore_error = False, retry_count = 3 sleep_time = 0.5 check_callback = at 0x7f1cb8be8040> parse = False def query_with_retry( self, sql, stdin=None, timeout=None, settings=None, user=None, password=None, database=None, host=None, ignore_error=False, retry_count=20, sleep_time=0.5, check_callback=lambda x: True, parse=False, ): # logging.debug(f"Executing query {sql} on {self.name}") result = None exception_msg = "" for i in range(retry_count): try: result = self.query( sql, stdin=stdin, timeout=timeout, settings=settings, user=user, password=password, database=database, host=host, ignore_error=ignore_error, parse=parse, ) if check_callback(result): return result time.sleep(sleep_time) except QueryRuntimeException as ex: exception_msg = f"{type(ex).__name__}: {str(ex)}" # Container is down, this is likely due to server crash. if "No route to host" in str(ex): raise time.sleep(sleep_time) except Exception as ex: # logging.debug("Retry {} got exception {}".format(i + 1, ex)) exception_msg = f"{type(ex).__name__}: {str(ex)}" time.sleep(sleep_time) if result is not None: return result > raise Exception(f"Can't execute query {sql}\n{exception_msg}") E Exception: Can't execute query E ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree() E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') E E QueryRuntimeException: Client failed! Return code: 198, stderr: Received exception from server (version 25.3.8): E Code: 198. DB::Exception: Received from 172.16.1.5:9000. DB::NetException. DB::NetException: Not found address of host: raw.githubusercontent.com: while loading disk metadata. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:113: Poco::Exception::Exception(String const&, int) @ 0x000000001f81eb40 E 1. ./build_docker/./src/Common/Exception.cpp:108: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000ff71534 E 2. ./src/Common/Exception.h:112: DB::NetException::NetException(int, FormatStringHelperImpl::type>, String const&) @ 0x000000000ff2de19 E 3. ./build_docker/./src/Common/DNSResolver.cpp:113: DB::(anonymous namespace)::hostByName(String const&) @ 0x000000000ff29783 E 4. ./build_docker/./src/Common/DNSResolver.cpp:138: DB::DNSResolver::getResolvedIPAdressessWithFiltering(String const&) @ 0x000000000ff27d53 E 5. ./build_docker/./src/Common/DNSResolver.cpp:256: DB::DNSResolver::resolveIPAddressWithCache(String const&) @ 0x000000000ff28339 E 6. ./build_docker/./src/Common/DNSResolver.cpp:276: DB::DNSResolver::resolveHostAllInOriginOrder(String const&) @ 0x000000000ff28a70 E 7. ./build_docker/./src/Common/HostResolvePool.cpp:54: std::vector> std::__function::__policy_invoker> (String const&)>::__call_impl[abi:ne190107]> (String const&)>>(std::__function::__policy_storage const*, String const&) @ 0x00000000103fb829 E 8. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x00000000103f968f E 9. ./build_docker/./src/Common/HostResolvePool.cpp:66: DB::HostResolver::HostResolver(std::function> (String const&)>&&, String, Poco::Timespan) @ 0x00000000103f949f E 10. ./build_docker/./src/Common/HostResolvePool.cpp:53: DB::HostResolver::HostResolver(String, Poco::Timespan) @ 0x00000000103f913e E 11. ./src/Common/HostResolvePool.h:62: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler::make_shared_enabler(String const&) @ 0x00000000103fe723 E 12. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler> std::allocate_shared[abi:ne190107] DB::HostResolver::create(String const&)::make_shared_enabler, std::allocator DB::HostResolver::create(String const&)::make_shared_enabler>, String const&, 0>(std::allocator DB::HostResolver::create(String const&)::make_shared_enabler> const&, String const&) @ 0x00000000103fe426 E 13. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:851: DB::HostResolversPool::getResolver(String const&) @ 0x00000000103fb6f1 E 14. ./build_docker/./src/Common/HTTPConnectionPool.cpp:671: DB::EndpointConnectionPool::prepareNewConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103edbb6 E 15. ./build_docker/./src/Common/HTTPConnectionPool.cpp:590: DB::EndpointConnectionPool::getConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103ece18 E 16. ./build_docker/./src/IO/HTTPCommon.cpp:63: DB::makeHTTPSession(DB::HTTPConnectionGroupType, Poco::URI const&, DB::ConnectionTimeouts const&, DB::ProxyConfiguration const&, unsigned long*) @ 0x000000001040592d E 17. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:267: DB::ReadWriteBufferFromHTTP::callImpl(Poco::Net::HTTPResponse&, String const&, std::optional const&, bool) const @ 0x0000000013133465 E 18. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:285: DB::ReadWriteBufferFromHTTP::callWithRedirects(Poco::Net::HTTPResponse&, String const&, std::optional const&) @ 0x0000000013133714 E 19. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:408: DB::ReadWriteBufferFromHTTP::initialize() @ 0x00000000131341c0 E 20. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:472: void std::__function::__policy_invoker::__call_impl[abi:ne190107]>(std::__function::__policy_storage const*) @ 0x0000000013137377 E 21. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x0000000013130bd9 E 22. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:465: DB::ReadWriteBufferFromHTTP::nextImpl() @ 0x0000000013135881 E 23. DB::ReadBuffer::next() @ 0x0000000007d5819b E 24. ./src/IO/ReadBuffer.h:96: DB::WebObjectStorage::loadFiles(String const&, std::unique_lock const&) const @ 0x0000000016d85459 E 25. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:225: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d888da E 26. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:185: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d882fa E 27. ./build_docker/./src/Disks/ObjectStorages/Web/MetadataStorageFromStaticFilesWebServer.cpp:106: DB::MetadataStorageFromStaticFilesWebServer::getStorageObjectsIfExist(String const&) const @ 0x0000000016d828a5 E 28. ./build_docker/./src/Disks/ObjectStorages/DiskObjectStorage.cpp:785: DB::DiskObjectStorage::readFileIfExists(String const&, DB::ReadSettings const&, std::optional, std::optional) const @ 0x0000000016ce8e92 E 29. ./build_docker/./src/Storages/MergeTree/MergeTreeData.cpp:380: DB::MergeTreeData::initializeDirectoriesAndFormatVersion(String const&, bool, String const&, bool) @ 0x000000001a7120fd E 30. ./build_docker/./src/Storages/StorageMergeTree.cpp:159: DB::StorageMergeTree::StorageMergeTree(DB::StorageID const&, String const&, DB::StorageInMemoryMetadata const&, DB::LoadingStrictnessLevel, std::shared_ptr, String const&, DB::MergeTreeData::MergingParams const&, std::unique_ptr>) @ 0x000000001abe70e7 E 31. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr std::allocate_shared[abi:ne190107], DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>, 0>(std::allocator const&, DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>&&) @ 0x000000001abe6a04 E . (DNS_ERROR) E (query: ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree() E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') E ) helpers/cluster.py:3722: Exception ------------------------------ Captured log call ------------------------------- 2026-04-10 18:40:21.042000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3658, query) 2026-04-10 18:40:21.258000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3658, query) 2026-04-10 18:40:21.474000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3658, query) 2026-04-10 18:40:21.741000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3658, query) 2026-04-10 18:40:21.957000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) 2026-04-10 18:41:16.657000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) 2026-04-10 18:42:13.714000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) _______________________ test_only_destination_replicated _______________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_only_destination_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:163: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = sql = "\n ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7'\n (\n price UInt32,\n ...disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/')\n " stdin = None, timeout = 60, settings = None, user = None, password = None database = None, host = None, ignore_error = False, retry_count = 3 sleep_time = 0.5 check_callback = at 0x7f1cb8be8040> parse = False def query_with_retry( self, sql, stdin=None, timeout=None, settings=None, user=None, password=None, database=None, host=None, ignore_error=False, retry_count=20, sleep_time=0.5, check_callback=lambda x: True, parse=False, ): # logging.debug(f"Executing query {sql} on {self.name}") result = None exception_msg = "" for i in range(retry_count): try: result = self.query( sql, stdin=stdin, timeout=timeout, settings=settings, user=user, password=password, database=database, host=host, ignore_error=ignore_error, parse=parse, ) if check_callback(result): return result time.sleep(sleep_time) except QueryRuntimeException as ex: exception_msg = f"{type(ex).__name__}: {str(ex)}" # Container is down, this is likely due to server crash. if "No route to host" in str(ex): raise time.sleep(sleep_time) except Exception as ex: # logging.debug("Retry {} got exception {}".format(i + 1, ex)) exception_msg = f"{type(ex).__name__}: {str(ex)}" time.sleep(sleep_time) if result is not None: return result > raise Exception(f"Can't execute query {sql}\n{exception_msg}") E Exception: Can't execute query E ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree() E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') E E QueryRuntimeException: Client failed! Return code: 198, stderr: Received exception from server (version 25.3.8): E Code: 198. DB::Exception: Received from 172.16.1.5:9000. DB::NetException. DB::NetException: Not found address of host: raw.githubusercontent.com: while loading disk metadata. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:113: Poco::Exception::Exception(String const&, int) @ 0x000000001f81eb40 E 1. ./build_docker/./src/Common/Exception.cpp:108: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000ff71534 E 2. ./src/Common/Exception.h:112: DB::NetException::NetException(int, FormatStringHelperImpl::type>, String const&) @ 0x000000000ff2de19 E 3. ./build_docker/./src/Common/DNSResolver.cpp:113: DB::(anonymous namespace)::hostByName(String const&) @ 0x000000000ff29783 E 4. ./build_docker/./src/Common/DNSResolver.cpp:138: DB::DNSResolver::getResolvedIPAdressessWithFiltering(String const&) @ 0x000000000ff27d53 E 5. ./build_docker/./src/Common/DNSResolver.cpp:256: DB::DNSResolver::resolveIPAddressWithCache(String const&) @ 0x000000000ff28339 E 6. ./build_docker/./src/Common/DNSResolver.cpp:276: DB::DNSResolver::resolveHostAllInOriginOrder(String const&) @ 0x000000000ff28a70 E 7. ./build_docker/./src/Common/HostResolvePool.cpp:54: std::vector> std::__function::__policy_invoker> (String const&)>::__call_impl[abi:ne190107]> (String const&)>>(std::__function::__policy_storage const*, String const&) @ 0x00000000103fb829 E 8. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x00000000103f968f E 9. ./build_docker/./src/Common/HostResolvePool.cpp:66: DB::HostResolver::HostResolver(std::function> (String const&)>&&, String, Poco::Timespan) @ 0x00000000103f949f E 10. ./build_docker/./src/Common/HostResolvePool.cpp:53: DB::HostResolver::HostResolver(String, Poco::Timespan) @ 0x00000000103f913e E 11. ./src/Common/HostResolvePool.h:62: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler::make_shared_enabler(String const&) @ 0x00000000103fe723 E 12. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr DB::HostResolver::create(String const&)::make_shared_enabler> std::allocate_shared[abi:ne190107] DB::HostResolver::create(String const&)::make_shared_enabler, std::allocator DB::HostResolver::create(String const&)::make_shared_enabler>, String const&, 0>(std::allocator DB::HostResolver::create(String const&)::make_shared_enabler> const&, String const&) @ 0x00000000103fe426 E 13. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:851: DB::HostResolversPool::getResolver(String const&) @ 0x00000000103fb6f1 E 14. ./build_docker/./src/Common/HTTPConnectionPool.cpp:671: DB::EndpointConnectionPool::prepareNewConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103edbb6 E 15. ./build_docker/./src/Common/HTTPConnectionPool.cpp:590: DB::EndpointConnectionPool::getConnection(DB::ConnectionTimeouts const&, unsigned long*) @ 0x00000000103ece18 E 16. ./build_docker/./src/IO/HTTPCommon.cpp:63: DB::makeHTTPSession(DB::HTTPConnectionGroupType, Poco::URI const&, DB::ConnectionTimeouts const&, DB::ProxyConfiguration const&, unsigned long*) @ 0x000000001040592d E 17. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:267: DB::ReadWriteBufferFromHTTP::callImpl(Poco::Net::HTTPResponse&, String const&, std::optional const&, bool) const @ 0x0000000013133465 E 18. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:285: DB::ReadWriteBufferFromHTTP::callWithRedirects(Poco::Net::HTTPResponse&, String const&, std::optional const&) @ 0x0000000013133714 E 19. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:408: DB::ReadWriteBufferFromHTTP::initialize() @ 0x00000000131341c0 E 20. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:472: void std::__function::__policy_invoker::__call_impl[abi:ne190107]>(std::__function::__policy_storage const*) @ 0x0000000013137377 E 21. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x0000000013130bd9 E 22. ./build_docker/./src/IO/ReadWriteBufferFromHTTP.cpp:465: DB::ReadWriteBufferFromHTTP::nextImpl() @ 0x0000000013135881 E 23. DB::ReadBuffer::next() @ 0x0000000007d5819b E 24. ./src/IO/ReadBuffer.h:96: DB::WebObjectStorage::loadFiles(String const&, std::unique_lock const&) const @ 0x0000000016d85459 E 25. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:225: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d888da E 26. ./build_docker/./src/Disks/ObjectStorages/Web/WebObjectStorage.cpp:185: DB::WebObjectStorage::tryGetFileInfo(String const&) const @ 0x0000000016d882fa E 27. ./build_docker/./src/Disks/ObjectStorages/Web/MetadataStorageFromStaticFilesWebServer.cpp:106: DB::MetadataStorageFromStaticFilesWebServer::getStorageObjectsIfExist(String const&) const @ 0x0000000016d828a5 E 28. ./build_docker/./src/Disks/ObjectStorages/DiskObjectStorage.cpp:785: DB::DiskObjectStorage::readFileIfExists(String const&, DB::ReadSettings const&, std::optional, std::optional) const @ 0x0000000016ce8e92 E 29. ./build_docker/./src/Storages/MergeTree/MergeTreeData.cpp:380: DB::MergeTreeData::initializeDirectoriesAndFormatVersion(String const&, bool, String const&, bool) @ 0x000000001a7120fd E 30. ./build_docker/./src/Storages/StorageMergeTree.cpp:159: DB::StorageMergeTree::StorageMergeTree(DB::StorageID const&, String const&, DB::StorageInMemoryMetadata const&, DB::LoadingStrictnessLevel, std::shared_ptr, String const&, DB::MergeTreeData::MergingParams const&, std::unique_ptr>) @ 0x000000001abe70e7 E 31. ./contrib/llvm-project/libcxx/include/__memory/construct_at.h:41: std::shared_ptr std::allocate_shared[abi:ne190107], DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>, 0>(std::allocator const&, DB::StorageID const&, String const&, DB::StorageInMemoryMetadata&, DB::LoadingStrictnessLevel const&, std::shared_ptr&, String&, DB::MergeTreeData::MergingParams&, std::unique_ptr>&&) @ 0x000000001abe6a04 E . (DNS_ERROR) E (query: ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' E ( E price UInt32, E date Date, E postcode1 LowCardinality(String), E postcode2 LowCardinality(String), E type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), E is_new UInt8, E duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), E addr1 String, E addr2 String, E street LowCardinality(String), E locality LowCardinality(String), E town LowCardinality(String), E district LowCardinality(String), E county LowCardinality(String) E ) E ENGINE = MergeTree() E ORDER BY (postcode1, postcode2, addr1, addr2) E SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') E ) helpers/cluster.py:3722: Exception ------------------------------ Captured log call ------------------------------- 2026-04-10 18:43:11.606000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3658, query) 2026-04-10 18:43:11.821000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3658, query) 2026-04-10 18:43:12.087000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3658, query) 2026-04-10 18:43:12.303000 [ 622 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3658, query) 2026-04-10 18:43:12.519000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) 2026-04-10 18:44:06.412000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) 2026-04-10 18:45:01.670000 [ 622 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3658, query) ---------------------------- Captured log teardown ----------------------------- 2026-04-10 18:45:58.916000 [ 622 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/.env --project-name roottestattachpartitionusingcopy-gw0 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/docker-compose.yml stop --timeout 20] (cluster.py:121, run_and_check) 2026-04-10 18:46:05.747000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:05.748000 [ 622 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2026-04-10 18:46:05.757000 [ 622 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2026-04-10 18:46:05.766000 [ 622 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/.env --project-name roottestattachpartitionusingcopy-gw0 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1-gw0/replica2/docker-compose.yml down --volumes] (cluster.py:121, run_and_check) 2026-04-10 18:46:06.175000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:06.175000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:06.175000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:06.175000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Removing (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Removing (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica1-1 Removed (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-replica2-1 Removed (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Removing (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Removing (cluster.py:147, run_and_check) 2026-04-10 18:46:06.176000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:46:06.177000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Removing (cluster.py:147, run_and_check) 2026-04-10 18:46:06.177000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo1-1 Removed (cluster.py:147, run_and_check) 2026-04-10 18:46:06.177000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo2-1 Removed (cluster.py:147, run_and_check) 2026-04-10 18:46:06.177000 [ 622 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw0-zoo3-1 Removed (cluster.py:147, run_and_check) 2026-04-10 18:46:06.177000 [ 622 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw0_default Removing (cluster.py:147, run_and_check) 2026-04-10 18:46:06.177000 [ 622 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw0_default Removed (cluster.py:147, run_and_check) 2026-04-10 18:46:06.177000 [ 622 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 18:46:06.197000 [ 622 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:46:06.217000 [ 622 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:46:06.240000 [ 622 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:46:06.240000 [ 622 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 18:46:06.259000 [ 622 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 18:46:06.259000 [ 622 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw0 (cluster.py:889, cleanup) 2026-04-10 18:46:06.260000 [ 622 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 18:46:06.283000 [ 622 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 18:46:06.283000 [ 622 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 18:46:06.310000 [ 622 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 18:46:06.311000 [ 622 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 18:46:06.311000 [ 622 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 18:46:06.311000 [ 622 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:46:06.333000 [ 622 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:46:06.333000 [ 622 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) ----------------- generated report log file: parallel0_1.jsonl ----------------- ============================== slowest durations =============================== 170.49s call test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 169.76s call test_cow_policy/test.py::test_cow_policy[cow_policy_multi_volume] 169.30s call test_attach_partition_using_copy/test.py::test_both_mergetree 167.17s call test_attach_partition_using_copy/test.py::test_only_destination_replicated 164.62s call test_attach_partition_using_copy/test.py::test_all_replicated 164.03s call test_cow_policy/test.py::test_cow_policy[cow_policy_multi_disk] 18.82s setup test_attach_partition_using_copy/test.py::test_all_replicated 14.51s setup test_cow_policy/test.py::test_cow_policy[cow_policy_multi_disk] 7.42s teardown test_attach_partition_using_copy/test.py::test_only_destination_replicated 6.83s teardown test_cow_policy/test.py::test_cow_policy[cow_policy_multi_volume] 0.00s teardown test_attach_partition_using_copy/test.py::test_all_replicated 0.00s teardown test_cow_policy/test.py::test_cow_policy[cow_policy_multi_disk] 0.00s teardown test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s setup test_cow_policy/test.py::test_cow_policy[cow_policy_multi_volume] 0.00s setup test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s teardown test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s setup test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s setup test_attach_partition_using_copy/test.py::test_only_destination_replicated =========================== short test summary info ============================ FAILED test_cow_policy/test.py::test_cow_policy[cow_policy_multi_disk] - Exce... FAILED test_attach_partition_using_copy/test.py::test_all_replicated - Except... FAILED test_cow_policy/test.py::test_cow_policy[cow_policy_multi_volume] - Ex... FAILED test_attach_partition_using_copy/test.py::test_both_mergetree - Except... FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated ======================== 6 failed in 700.87s (0:11:40) ========================= Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 492, in subprocess.check_call(cmd, shell=True, bufsize=0) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_9p9r35 --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel0_1.jsonl --report-log-exclude-logs-on-passed-tests test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated 'test_cow_policy/test.py::test_cow_policy[cow_policy_multi_disk]' 'test_cow_policy/test.py::test_cow_policy[cow_policy_multi_volume]' -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 ' returned non-zero exit status 1.